science funding

=institutions =funding =research

 

 

Current science funding systems are widely considered unsatisfactory, but there's no consensus on what to do differently. Here are my thoughts on that topic.

 

 

some current issues

 

- Professors spend a lot of their time writing grants. In many cases, most of their time is taken up by grants and meetings.
- Grant distribution is slow.
- Grants are determined by average scores among a committee, so controversial proposals lose to more conventional ones.
- Grant reviewers want to fund the approach that seems most promising, and they don't know what other reviewers are approving, so redundant research on consensus approaches gets funded.
- In many departments in US universities, most graduate students are foreign nationals, and most of them leave after getting their degree.
- Journals charge high fees for access despite not paying reviewers, giving them enormous profits from the name and prestige of journals. Professors try to publish in those journals anyway, because promotions and grants are based partly on prestige of journals.
- Reviewers - being unpaid - often don't do a good job or take a long time.
- Publication is slow, partly because reviewers take a long time to respond.
- Faking research results is incentivized.
- Some reviewers vote against proposals then steal ideas from them.
- Some reviewers demand that their own papers be cited, because citations are another important metric.
- Some reviewers vote against research directions that could compete with their specialty.
- People trade citations for favors or other citations. This is now widespread.
- Because people who trade citations and otherwise exploit the system do better, they're selected for professorships.
- University labs tend to be little kingdoms that don't share resources or cooperate.

 

This is by no means meant to be a comprehensive list.

 

 

some existing proposals

 

Some papers have concluded that randomly distributing funding to proposals would work as well as the current system. However, if funding was distributed completely randomly, there would be more low-quality proposals. Most proposals have a minimum score threshold followed by a random lottery; that system has been used in practice, and works well enough, but this doesn't solve the above issues.

Some people are now proposing giving funds to professors who must redistribute those funds to other people's research but can do so freely. If this was used for most government research funding, it would result in large-scale corruption and patronage networks taking over research. Such networks take some time to develop, as citation rings did, but I expect any improvement would be temporary.

 

 

grant writing

 

Professors spend too much time writing grants. My proposal is:
All grant proposals for government money must go through a central system. Individuals are limited to 1 grant proposal per month. Professors can submit proposals for any students in their lab, but if they do so, funding is tied to that student, not the lab, and the student can go to another organization that has the necessary equipment and a history of acceptable results.

 

 

citations

 

I often look for information on a recent research topic, which leads to me often looking at minor scientific papers from the past 10 years. Sometimes I want to check for more-recent related results, so I'll search for papers citing what I read. It's common for the majority of those search results to be in the same field but not have any relation to the paper's topic. I suspect that most of those citations are traded or added as favors to eg reviewers or people in the same lab or university. If my perception is accurate, a large fraction of citations of papers in the past 10 years are not based on the paper's usefulness. Citation rings are a big deal, but I'm guessing that most questionable citations are traded less formally than that.

Many professors at US universities are already tenured based on traded citations, and those professors aren't likely to make the system work better. Many of the current stakeholders benefit from citation trading, so attempts at reform might only block competition to existing citation rings.

 

"Salami slicing" is another common practice. Instead of publishing one good paper, if you split it into minimal publishable units, you can get more publications and citations.

 

My proposal is:
Instead of using citation counts, make authors of papers list their top 3 citations, ranked. Governments would need to make authors do this for past papers, as well. Then, only count those 3 citations for evaluations.

 

That would:

- make weak and unrelated citations worthless
- make salami slicing counterproductive
- make it difficult for reviewers to demand useful self-citations

 

 

objective proxies

 

Students applying to universities generally take standardized multiple-choice tests, because scores on them are easy to measure objectively. If students trying to start an undergraduate or graduate program must take standardized tests, why don't professors have to as well?

One objection is that the standardized tests for students would be too easy. To get to a professorship, people generally did well on some standardized tests at some earlier stage. The resulting pool of graduate students is too large, especially if you include potential ones, but simply raising score thresholds isn't good, because the tests lose meaning at some point. In general, multiple-choice tests aren't very useful for distinguishing between test takers that are smarter than the test writers. For example, smart people taking IQ tests end up having to guess at the thought process of the test writers. So, good professor-level standardized tests aren't trivial.

A simple way to get more distinction between top scorers with a standardized test is to have more problems and make people solve them quickly. This is what IIT exams do. This selects for fast thinkers, and humans have a tradeoff between fast and deep thinking, so this selects against deep thinkers. (Someone probably wants to reply with a comment about John von Neumann, but fast mental math being a dumb person's idea of smart people is part of why von Neumann got credit for so many other people's ideas.)

Fortunately, there are already some more difficult standardized tests - for example, the International Math Olympiad is probably difficult enough to challenge many professors. However, most fields don't have something comparable, and math research is already one of the more meritocratic fields.

Another objection is that standardized tests aren't very similar to research or teaching, and more expensive evaluation than students get is worthwhile for professors. However, if the goal is to avoid quid pro quo outweighing merit, then some sort of system with automated and thus objective evaluation is probably necessary.

Harvard entrance exams used to test people on their knowledge of Latin and classical construction. Today, perhaps some video game would be suitable. But - what kind of video game?

Chess? No, chess is too *small*. Hiring researchers based on their skill at chess would be like hiring soccer players based on their rank in an arm wrestling tournament. It's better to recruit soccer players from sports with more well-rounded athletes, such as water polo or pentathlons.

On the other hand, when it comes to research, people need to specialize, so it's not good to rank people by their average score across many areas. It's probably better to let people choose a few things to be ranked on.

 

Some other things that aren't good for this:

- games largely about precise timing (eg Hades, Celeste)
- games largely about aiming at targets (eg FPS games)
- games that can largely be solved by memorization
- games that AI is better at than humans

 

The most suitable game types for this purpose seem to be:

- roguelikes
- builders with random challenges
- programming challenges

 

For example:

- leetcode
- DCSS
- Vault of the Void
- modded Kerbal Space Program, given a destination and time limit and set of usable parts, and minimizing initial cost
- Satisfactory, given a starting save and set of products to make and a time limit, and being scored on factory aesthetics by a panel of reviewers who look at save files
- Ludum Dare type competitions

 

One problem with both games and standardized tests as a metric for people is selection of and training for over-reliance on stated rules and systems working as intended. For example, in Kerbal Space Program, players often make complex staging arrangements that, if realistic system reliability was considered, would be bad. Ludum Dare as a challenge partly avoids this issue, especially if using Unity as a game engine.

 

 

Well, I'm talking a lot about video games and standardized tests here, but the point is just to have some sort of objective competition. Neural network research has been driven forwards largely by competitions such as ImageNet benchmarks. If something like computer chess competitions or protein folding or optimizing power inverters is relevant, then that's often better. The problem is, that's often not possible or very expensive. Computer game competitions have low capital costs and the bonus of being potentially entertaining to watch.

 

 


back to index